Tags: prompt engineering*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. This handbook provides a comprehensive introduction to Claude Code, Anthropic's AI-powered software development agent. It details how Claude Code differs from traditional autocomplete tools, functioning as an agent that reads, reasons about, and modifies codebases with user direction. The guide covers installation, initial setup, advanced workflows, integrations, and autonomous loops. It's aimed at developers, founders, and anyone seeking to leverage AI in software creation, emphasizing building real applications, accelerating feature development, and maintaining codebases efficiently. The handbook also highlights the importance of prompt discipline, planning, and understanding the underlying model to maximize Claude Code's capabilities.
  2. This article explains how to implement function calling with Google’s Gemma 3 27B model. It covers the concept of function calling, the step‑by‑step workflow, and provides a practical example using a Python `convert` function to turn $200,000 into EUR. The post walks through prompting Gemma, parsing its `tool_code` output, executing the function with `eval`, and returning a friendly final response. It also demonstrates how to set up the Google‑GenAI SDK, create a chat session, and extract tool calls. The discussion highlights Gemma’s multilingual, multimodal, and agentic capabilities, making it suitable for real‑world AI assistants that need to interact with external APIs and tools.
  3. Prompt caching significantly reduces LLM costs and latency by storing and reusing responses to repeated or similar prompts. The core technique involves checking a cache before sending a prompt to the LLM, retrieving a prior result if available. Effective caching requires balancing cache size, retrieval speed (using methods like vector databases), and strategies for handling slight prompt variations.
  4. This article discusses how to conduct long-term research effectively using AI as a partner, moving beyond single-prompt queries. It emphasizes the need for "Long-Term Triangulation" – a continuous, iterative methodology. The author outlines four key pillars: building a persistent memory for the AI, tracking shifts in the AI's understanding, actively critiquing its responses with contradictory data, and performing meta-audits to identify blind spots in the research process. The goal is to foster productive friction and avoid intellectual echo chambers, ensuring both the human and the AI think critically.
  5. Comprehensive guide to prompt engineering techniques for Claude's
    latest models, including Claude Opus 4.6, Claude Sonnet 4.6, and
    Claude Haiku 4.5. It covers foundational techniques, output
    control, tool use, thinking, and agentic systems.
  6. This article discusses how to effectively prompt local Large Language Models (LLMs) like those run with LM Studio or Ollama. It explains that local LLMs behave differently than cloud-based models and require more explicit and structured prompts for optimal results. The article provides guidance on how to craft better prompts, including using clear language, breaking down tasks into steps, and providing examples.
  7. An exploration of Claude 3 Opus's coding capabilities, specifically its ability to generate a functional CLI tool for the Minimax algorithm with a single prompt. The article details the prompt used, the generated code, and the successful execution of the tool, highlighting Claude's impressive one-shot learning and code generation abilities.
  8. Unusually detailed post explains how OpenAI handles the Codex agent loop. The article dives into the technical aspects of OpenAI's Codex CLI coding agent, including the agent loop, prompt construction, caching, and context window management.

    The article details how their Codex CLI coding agent functions. OpenAI engineer Michael Bolin explains the "agent loop" – the process by which the AI receives user input, generates code, runs tests, and iterates with human supervision.

    * **Agent Loop Mechanics:** The agent builds prompts with prioritized components (system, developer, user, assistant) and sends them to OpenAI’s Responses API.
    * **Prompt Management:** The system handles growing prompt lengths (quadratic growth) through caching, compaction, and a stateless API design (allowing for "Zero Data Retention"). Cache misses can significantly impact performance.
    * **Context Window:** Codex automatically compacts conversations to stay within the AI model's context window.
    * **Open Source Focus:** OpenAI open-sources the CLI client for Codex, unlike ChatGPT, suggesting a different approach to development and transparency for coding tools.
    * **Challenges Acknowledged:** The article doesn't shy away from the engineering challenges, like performance issues and bugs encountered during development.
    * **Future Coverage:** Bolin plans to release further posts detailing the CLI’s architecture, tool implementation, and sandboxing model.
  9. Repeating the input prompt improves performance for popular LLMs (Gemini, GPT, Claude, and Deepseek) without increasing the number of generated tokens or latency, when not using reasoning.
  10. Guides, papers, lessons, notebooks and resources for prompt engineering, context engineering, RAG, and AI Agents.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "prompt engineering"

About - Propulsed by SemanticScuttle